Medical Image Analysis
○ Elsevier BV
Preprints posted in the last 7 days, ranked by how well they match Medical Image Analysis's content profile, based on 33 papers previously published here. The average preprint has a 0.06% match score for this journal, so anything above that is already an above-average fit.
Kritopoulos, G.; Neofotistos, G.; Barmparis, G. D.; Tsironis, G. P.
Show abstract
Class imbalance in clinical electrocardiogram (ECG) datasets limits the diagnostic sensitivity of automated arrhythmia classifiers, particularly for rare but clinically significant beat types. We propose a three-stage hybrid generative pipeline that combines a spectral-guided conditional Variational Autoencoder (cVAE), a class-conditional latent Denoising Diffusion Probabilistic Model (DDPM), and a Quantum Latent Refinement (QLR) module built on parameterized quantum circuits to augment minority arrhythmia classes in the MIT-BIH Arrhythmia Database. The QLR module applies a bounded residual correction guided by Maximum Mean Discrepancy minimization to align synthetic latent distributions with real class-specific latent banks. A lightweight 1D MobileNetV2 classifier evaluated over five independent random seeds and four augmentation ratios serves as the downstream benchmark. Our findings establish latent diffusion augmentation as an effective strategy for imbalanced ECG classification and motivate further investigation of quantum-classical hybrid methods in cardiac diagnostics.
Hakata, Y.; Oikawa, M.; Fujisawa, S.
Show abstract
Background. Federated learning (FL) enables collaborative model training across institutions without sharing patient-level data. However, standard FL algorithms such as FedAvg degrade under non-independently and non-identically distributed (non-IID) data, a prevalent condition when patient demographics, scanner hardware, and disease prevalence differ across hospital sites. Objective. We propose iPS-MFFL (Individualized Per-Site Meta-Federated Feature Learning), a federated framework with a hierarchical local-model architecture that addresses non-IID heterogeneity through (1) a shared feature extractor, (2) multiple weak-learner classification heads that can be trained with heterogeneous training objectives to promote complementary decision boundaries, (3) independent per-learner server aggregation so that each weak learner's parameters are averaged only with its counterparts at other clients, and (4) a lightweight meta-model, itself federated, that adaptively stacks the weak-learner outputs. Methods. We evaluate on the Brain Tumor MRI Classification dataset (7,200 images; 4 classes: glioma, meningioma, pituitary tumor, no tumor) partitioned across K = 5 simulated hospital sites using Dirichlet non-IID sampling (alpha = 0.3). Four baselines are compared: Local-only training, FedAvg, FedProx, and Freeze-FT. All experiments are repeated over three random seeds (13, 42, 2025) and evaluated using paired t-tests, Cohen's d effect sizes, and post-hoc power analysis.
Hakata, Y.; Oikawa, M.; Fujisawa, S.
Show abstract
Background. Adult diffuse glioma is a representative class of primary brain tumors for which accurate MRI-based tumor segmentation is indispensable for treatment planning. Conventional automated segmentation methods have relied primarily on image information and spatial prompts, and auxiliary clinical information that is routinely acquired in clinical practice has not been sufficiently exploited as an input. Objective. Building on a dual-prompt-driven Segment Anything Model (SAM) extension framework that fuses visual and language reference prompts, we propose a method that integrates patient demographics, unsupervised molecular cluster variables derived from TCGA high-throughput profiling, and histopathological parameters as learnable prompt embeddings, and we evaluate its effect on the accuracy of lower-grade glioma (LGG) MRI segmentation. Methods. An auxiliary prompt encoder converts clinical metadata into high-dimensional embeddings that are fused with the prompt representations of Segment Anything Model (SAM) ViT-B through a cross-attention fusion mechanism. The TCGA-LGG MRI Segmentation dataset (Kaggle release by Buda et al.; n = 110 patients; WHO grade II-III) was split at the patient level (train/val/test = 71/17/22) using three different random seeds, and the three slices with the largest tumor area were extracted from each patient. To avoid pseudo-replication arising from multiple slices per patient and repeated measurements across seeds, our primary analysis aggregated Dice and 95th-percentile Hausdorff distance (HD95) to the patient x seed unit (n = 66); secondary analyses at the unique-patient level (n = 22) and at the per-slice level (n = 198) are also reported. Pairwise comparisons used paired t-tests with Bonferroni correction (k = 3) and Wilcoxon signed-rank tests, and a permutation test (K = 30) served as an auxiliary check of effective use of the auxiliary information. Results. At the patient x seed level (n = 66), Proposed (full clinical) achieved a Dice gain of +0.287 over the zero-shot SAM ViT-B baseline (paired-t p = 4.2 x 10^-15, Cohen's d_z = +1.25, Bonferroni-corrected p << 0.001; Wilcoxon p = 2.0 x 10^-10), and HD95 improved from 218.2 to 64.6. Because zero-shot SAM is not designed for domain-specific medical segmentation, the large absolute HD95 gap largely reflects the expected domain gap rather than a competitive baseline. The additional contribution of the full clinical configuration over the demographics-only configuration was Dice = +0.023 (paired-t p = 0.057, Bonferroni-corrected p = 0.172), which did not reach statistical significance at the patient level and is reported as a directional trend. The permutation test (K = 30, seed 2025) yielded real-metadata Dice = 0.819 versus a shuffled-metadata mean of 0.773, giving an empirical p = 0.032 = 1/(K + 1), which is at the resolution limit of this test and should therefore be interpreted as preliminary evidence. Conclusions. Integrating auxiliary clinical information as multimodal prompts produced a large improvement over the zero-shot SAM baseline on this LGG cohort. More importantly, a robustness analysis showed that Proposed (full clinical) outperformed the trained Base (no auxiliary information) under all tested spatial-prompt conditions, including perfect centroid (+0.014), and that the advantage was most pronounced in the prompt-free regime (+0.231, p = 0.039), where the base model collapsed but the proposed model maintained meaningful segmentation by leveraging clinical metadata alone. The additional contribution of molecular and histopathological information beyond demographics was not statistically resolved at the patient level (+0.023, n.s.). Establishing clinical utility will require external validation on larger multi-center cohorts and direct comparisons with established segmentation methods. Keywords: brain tumor segmentation; Segment Anything Model (SAM); vision-language prompt-driven segmentation; auxiliary clinical prompts; multimodal learning; TCGA-LGG; deep learning
Roca, M.; Messuti, G.; Klepachevskyi, D.; Angiolelli, M.; Bonavita, S.; Trojsi, F.; Demuru, M.; Troisi Lopez, E.; Chevallier, S.; Yger, F.; Saudargiene, A.; Sorrentino, P.; Corsi, M.-C.
Show abstract
Neurodegenerative diseases such as Mild Cognitive Impairment (MCI), Multiple Sclerosis (MS), Parkinson s Disease (PD), and Amyotrophic Lateral Sclerosis (ALS) are becoming more prevalent. Each of these diseases, despite its specific pathophysiological mechanisms, leads to widespread reorganization of brain activity. However, the corresponding neurophysiological signatures of these changes have been elusive. As a consequence, to date, it is not possible to effectively distinguish these diseases from neurophysiological data alone. This work uses Magnetoencephalography (MEG) resting-state data, combined with interpretable machine learning techniques, to support differential diagnosis. We expand on previous work and design a Riemannian geometry-based classification pipeline. The pipeline is fed with typical connectivity metrics, such as covariance or correlation matrices. To maintain interpretability while reducing feature dimensionality, we introduce a classifier-independent feature selection procedure that uses effect sizes derived from the Kruskal-Wallis test. The ensemble classification pipeline, called REDDI, achieved a mean balanced accuracy of 0.81 (+/-0.04) across five folds, representing a 13% improvement over the state-of-the-art, while remaining clinically transparent. As such, our approach achieves reliable, interpretable, data-driven, operator-independent decision-support tools in Neurology.
Wang, S.; Ayubcha, C.; Hua, Y.; Beam, A.
Show abstract
Background: Developing generalizable neuroimaging models is often hindered by limited labeled data which has led to an increased interest in unsupervised inverse learning. Existing approaches often neglect geometric principles and struggle with diverse pathologies. We propose a symmetry-informed inverse learning foundation model to address these shortcomings for robust and efficient anomaly detection in brain MRI. Methods: Our framework employs a reconstruction-to-embedding pipeline, trained exclusively on healthy brain MRI slices. A 2D U-Net uses a novel, symmetry-aware masking strategy to reconstruct a disorder-free slice. Difference maps are embedded into a 1024-dimensional latent space via a Beta-VAE. Anomaly scoring is performed using Mahalanobis distance. We evaluated generalization by fine-tuning on external lesion datasets, BraTS Africa (SSA), and the ADNI-derived Alzheimer disease cohort (Alz). Results: On the source metastasis (Mets) dataset, the framework achieved high performance (AB1+MSE: 99.28% accuracy, 99.79% sensitivity). Generalization to the external lesion dataset (SSA) was robust, with the Symmetry ROC configuration achieving 91.93% accuracy. Transfer to the Alzheimer dataset (Alz) was more challenging, achieving a peak accuracy of 70.54% with a high false-positive rate, suggesting difficulty in separating subtle, diffuse changes. Conclusion: The symmetry-informed inverse learning framework establishes a robust foundation model for neuroimaging, showing strong performance for focal lesions and successful generalization under domain shift. Limitations in diffuse neurodegeneration underscore the necessity for richer representations and multimodal integration to improve future foundation models.
Zhou, J.; Miller, R. J.; Shanbhag, A.; Killekar, A.; Han, D.; Patel, K. K.; Pieszko, K.; Yi, J.; Urs, M. K.; Ramirez, G.; Lemley, M.; Kavanagh, P. B.; Liang, J. X.; Kamagate, A.; Builoff, V.; Einstein, A. J.; Feher, A.; Miller, E. J.; Sinusas, A. J.; Ruddy, T. D.; Knight, S.; Le, V. T.; Mason, S.; Chareonthaitawee, P.; Wopperer, S.; Alexanderson, E.; Carvajal-Juarez, I.; Rosamond, T. L.; Slipczuk, L.; Travin, M. I.; Packard, R. R.; Acampa, W.; Al-Mallah, M.; deKemp, R. A.; Buechel, R. R.; Berman, D. S.; Dey, D.; Di Carli, M. F.; Slomka, P. J.
Show abstract
Purpose: Spatial distribution of coronary artery calcium (CAC) may provide additional prognostic value in patients undergoing SPECT and PET myocardial perfusion imaging (MPI). We aimed to automatically identify CAC in proximal segments from attenuation correction CT (CTAC) scans using artificial intelligence (AI) and to evaluate prognostic significance in two large international multicenter registries. Methods: From hybrid MPI/CT imaging (N=43,099) across 15 sites, we included 4,552 most relevant patients with 1) no prior coronary artery disease; 2) AI-derived mild CAC scores (1-99); and 3) normal perfusion (stress total perfusion deficit <5%). The independent associations between AI-identified proximal CAC and major adverse cardiovascular events (MACE) and all-cause mortality (ACM) were evaluated using multivariable Cox regression, likelihood ratio test (LRT), and continuous net reclassification index (NRI). Results: Among the patients with mild CAC and normal perfusion (mean age 65{+/-}12 years, 51% male), 1,730 (38%) had proximal CAC. Over 3.6 (inter-quartile interval 2.1, 5.2) years follow up, 599 (13%) and 444 (10%) patients had MACE or ACM, respectively. Proximal CAC was associated with an increased risk of MACE (adjusted hazard ratio [HR] 1.24, 95% CI 1.03-1.48, P=0.02) and ACM (adjusted HR 1.25, 95% CI 1.01-1.53, P=0.04) after the adjustment of CAC score and density, clinical risk factors, and perfusion deficit. Proximal CAC improved the risk stratification of MACE (LRT P=0.02; NRI 12%) and ACM (LRT P=0.04; NRI 12%). Conclusion: In patients with mild CAC and normal perfusion, AI detection of proximal CAC identified a higher-risk group for adverse outcomes, highlighting its prognostic utility.
Nguyen, T. M.; Woods, C.; Liu, J.; Wang, C.; Lin, A.-L.; Cheng, J.
Show abstract
The apolipoprotein E {varepsilon}4 (APOE4) allele is the strongest genetic risk factor for late-onset Alzheimer's disease (AD), the most common form of dementia. APOE4 carriers exhibit cerebrovascular and metabolic dysfunction, structural brain alterations, and gut microbiome changes decades before the onset of clinical symptoms. A better understanding of the early manifestation of these physiological changes is critical for the development of timely AD interventions and risk reduction protocols. Multimodal datasets encompassing a wide range of APOE4- and AD-associated biomarkers provide a valuable opportunity to gain insight into the APOE4 phenotype; however, these datasets often present analytical challenges due to small sample sizes and high heterogeneity. Here, we propose a two-stage multimodal AI model (APOEFormer) that integrates blood metabolites, brain vascular and structural MRI, microbiome profiles, and other clinical and demographic data to predict APOE4 allele status. In the first stage, modality-specific encoders are used to generate initial representations of input data modalities, which are aligned in a shared latent space via self-supervised contrastive learning during pretraining. This objective encourages the learning of informative and consistent representations across modalities by leveraging cross-modality relationships. In the second stage, the pretrained representations are used as inputs to a multimodal transformer that integrates information across modalities to predict a key AD risk genetic variant (APOE4). Across 10 independent experimental runs with different train-validation-test splits, APOEFormer predicts whether an individual carries an APOE4 allele with an average accuracy of 75%, demonstrating robust performance under limited sample sizes. Post hoc perturbation analysis of the predictive model revealed valuable insights into the driving components of the APOE4 phenotype, including key blood biomarkers and brain regions strongly associated with APOE4.
Aquaro, G. D.; Licordari, R.; De Gori, C.; Todiere, G.; Ianni, U.; Barison, A.; De Luca, A.; Folgheraiter, a.; Grigoratos, C.; alberti, m.; lombardo, m.; De Caterina, R.; Sinagra, G.; Emdin, M.; Di Bella, G.; fulceri, l.
Show abstract
Background: Late gadolinium enhancement (LGE) quantification by cardiovascular magnetic resonance is central to risk stratification in hypertrophic cardiomyopathy (HCM), yet conventional techniques require contour tracing and region-of-interest (ROI) placement, which may reduce reproducibility and increase analysis time. We developed a novel visual standardized approach, the Visual Standardized Quantification of LGE (VISTAQ), that does not require myocardial contouring, arbitrary ROI positioning, or dedicated post-processing software. Methods: In this multicenter, multivendor retrospective study, LGE images from 400 patients (100 prior myocardial infarction, 250 HCM, 50 other non-ischemic heart diseases) were analyzed. VISTAQ subdivides each myocardial segment into transmural mini-segments and classifies LGE visually using predefined criteria, expressing global LGE burden as the percentage of positive mini-segments. Reproducibility was assessed in 250 patients across different observer expertise levels using intraclass correlation coefficients (ICC) and Bland?Altman analysis. In 100 HCM patients, VISTAQ was compared with conventional methods (mean+2SD, +5SD, +6SD, FWHM, visual thresholding). Prognostic performance was evaluated in 250 HCM patients over a median 5-year follow-up. Results: VISTAQ demonstrated excellent intra- and inter-observer reproducibility (ICC up to 0.98 and 0.97, respectively), consistent across disease subtypes. Compared with conventional techniques, VISTAQ showed similar ICC to FWHM but significantly lower net and absolute inter-observer differences (median absolute difference 1.3%). Mean+2SD markedly overestimated LGE, whereas mean+6SD slightly underestimated LGE compared with VISTAQ, mean+5SD, FWHM, and visual thresholding. Analysis time was substantially shorter with VISTAQ (median 105 vs. 375 seconds, p<0.0001). During follow-up, 21 hard cardiac events occurred in HCM population. An LGE threshold >10% predicted events with higher accuracy using VISTAQ (AUC 0.90; sensitivity 85%; specificity 94%) compared with mean+6SD (AUC 0.75; sensitivity 57%; specificity 93%). Conclusions: VISTAQ provides highly reproducible, time-efficient LGE quantification without dedicated software and demonstrates non-inferior prognostic discrimination in HCM compared with conventional threshold-based techniques.
Zhang, Q.; Tang, Q.; Vu, T.; Pandit, K.; Cui, Y.; Yan, F.; Wang, N.; Li, J.; Yao, A.; Menozzi, L.; Fung, K.-M.; Yu, Z.; Parrack, P.; Ali, W.; Liu, R.; Wang, C.; Liu, J.; Hostetler, C. A.; Milam, A. N.; Nave, B.; Squires, R. A.; Battula, N. R.; Pan, C.; Martins, P. N.; Yao, J.
Show abstract
End-stage liver disease (ESLD) is one of the leading causes of death worldwide. Currently, the only curative option for patients with ESLD is liver transplantation. However, the demand for donor livers far exceeds the available supply, partly because many potentially viable livers are discarded following biopsy evaluation. While biopsy is the gold standard for assessing liver histological features related to graft quality and transplant suitability, it often leads to high discard rates due to its susceptibility to sampling errors and limited spatial coverage. Besides, biopsy is invasive, time-consuming, and unavailable in clinical facilities with limited resources. Here, we present an AI-assisted photoacoustic/ultrasound (PA/US) imaging framework for quantitative assessment of human donor liver graft quality and transplant suitablity at the whole-organ scale. With multimodal volumetric PA/US images as the input, our deep-learning (DL) model accurately predicted the risk level of fibrosis and steatosis, which indicate the graft quality and transplant suitability, when comparing with true pathological scores. DL also identified the imaging modes (PAI wavelength and B-mode USI) that correlated the most with prediction accuracy, without relying on ill-posed spectral unmixing. Our method was evaluated in six discarded human donor livers comprising sixty spatially matched regions of interest. Our study will pave the way for a new standard of care in organ graft quality and transplant suitability that is fast, noninvasive, and spatially thorough to prevent unnecessary organ discards in liver transplantation.
Chandra, S.
Show abstract
Background: Current deep learning models in computational pathology, radiology, and digital pathology produce opaque predictions that lack the explainable artificial intelligence (xAI) capabilities required for clinical adoption. Despite achieving radiologist-level performance in tasks from whole-slide image (WSI) classification to mammographic screening, these models function as black boxes: clinicians cannot trace predictions to specific biological features, verify outputs against established morphological criteria, or integrate AI reasoning into precision oncology workflows and tumor board decision-making. Methods: We present Virtual Spectral Decomposition (VSD), a modality-agnostic, interpretable-by-design framework that decomposes medical images into six biologically interpretable tissue composition channels using sigmoid threshold functions - the same mathematical structure as CT windowing. Unlike post-hoc xAI methods (Grad-CAM, SHAP, LIME) applied to black-box deep learning models, VSD channels have pre-defined biological meanings derived from tissue physics, providing inherent explainability without sacrificing quantitative rigor. For whole-slide image (WSI) analysis in digital pathology, we introduce the dendritic tile selection algorithm, a biologically-inspired hierarchical architecture achieving 70-80% computational reduction while preferentially sampling the tumor immune microenvironment. VSD is validated across three cancer types and imaging modalities: pancreatic ductal adenocarcinoma (PDAC) on CT imaging, lung adenocarcinoma (LUAD) on H&E-stained pathology slides using TCGA data, and breast cancer on screening mammography. Composition entropy of the six-channel vector is computed as a visual Biological Entropy Index (vBEI) - an imaging biomarker quantifying the diversity of active biological defense systems. Results: In pancreatic cancer, the fat-to-stroma ratio (a novel CT-derived radiomics biomarker) declines from >5.0 (normal) to <0.5 (advanced PDAC), enabling early detection of desmoplastic invasion before mass formation on standard imaging. In lung cancer, composition entropy from H&E whole-slide images correlates with tumor immune microenvironment markers from RNA-seq (CD3: rho=+0.57, p=0.009; CD8: rho=+0.54, p=0.015; PD-1: rho=+0.54, p=0.013) and predicts overall survival (low entropy immune-desert phenotype: 71% mortality vs 29%, p=0.032; n=20 TCGA-LUAD), providing immune phenotyping for checkpoint immunotherapy patient selection from a $5 H&E slide without molecular assays. In breast cancer, each lesion type produces a characteristic six-channel fingerprint functioning as an interpretable computer-aided diagnosis (CAD) system for quantitative BI-RADS assessment and subtype classification (IDC vs ILC vs DCIS vs IBC). A five-level xAI audit trail provides complete traceability from clinical decision support output to specific biological structures visible on the original images. Conclusion: VSD establishes a unified, interpretable-by-design mathematical framework for explainable tissue composition analysis across imaging modalities and cancer types. Unlike black-box deep learning and post-hoc xAI approaches, VSD provides inherently interpretable, clinically verifiable cancer detection and immune phenotyping from standard clinical imaging at existing costs - without requiring foundation model infrastructure, specialized hardware, or molecular assays. The open-source pipeline (Google Colab, Supplementary Material) enables immediate reproducibility and extension to additional cancer types across the pan-cancer TCGA atlas.
Wang, X.-Y.; Li, M.-M.; Zhao, S.-M.; Jia, X.-Y.; Yang, W.-S.; Chang, L.-L.; Wang, H.-M.; Zhao, J.-T.
Show abstract
Stroke-associated pneumonia (SAP) is a common, severe complication in acute ischemic stroke (AIS) patients receiving bridging therapy (intravenous thrombolysis + mechanical thrombectomy), worsening prognosis and increasing mortality. Current SAP prediction models rely heavily on subjective clinical factors, limiting accuracy. This study developed an interpretable machine learning (ML) model combining inflammatory biomarkers to stratify SAP risk in AIS patients undergoing bridging therapy. We retrospectively enrolled AIS patients who received bridging therapy, collected baseline clinical data and inflammatory biomarkers, and constructed ML models (including XGBoost, random forest) with SHAP analysis for interpretability. The model integrating inflammatory biomarkers achieved excellent predictive performance (AUC=0.XX, 95%CI: XX-XX), outperforming traditional clinical models. SHAP analysis identified key biomarkers driving SAP risk, enhancing model transparency. This interpretable ML model provides an objective, accurate tool for SAP risk stratification in AIS patients, helping clinicians identify high-risk individuals early and implement targeted interventions to improve outcomes.
Vikström, A.; Zarrinkoob, L.; Johannesdottir, M.; Wahlin, A.; Hellström, J.; Appelblad, M.; Holmlund, P.
Show abstract
Modelling of hemodynamics in the circle of Willis (CoW) depends on vascular segmentation, which may vary based on imaging modality. Computed tomography angiography (CTA) is commonly used in clinic but involves radiation and injection of contrast agents, whereas magnetic resonance angiography (MRA) offers a non-invasive alternative. This study aims to compare CoW morphology and modelled cerebral perfusion pressure of CTA and MRA segmentations, validating if MRA can replace CTA in modelling workflows. CTA and time-of-flight MRA (TOF-MRA) of the CoW was performed in 19 patients undergoing elective aortic arch surgery (67{+/-}7 years, 8 women). The CoW was semi-automatically segmented based on signal intensity thresholding. A TOF-MRA threshold was optimized against the CTA segmentation, using the CTA as reference standard. Computational fluid dynamics (CFD) modelling with boundary conditions based on subject-specific flow rates from 4D flow MRI simulated cerebral perfusion pressure in the segmented geometries. A baseline simulation and a unilateral brain inflow simulation, i.e., occlusion of a carotid, were carried out. Linear mixed models indicated there was no effect of choice of modality on either average arterial lumen area (CTA - TOF-MRA: -0.2{+/-}1.3 mm2; p=0.762) or baseline pressure drops (0.2{+/-}1.9 mmHg; p=0.257). In the unilateral inflow simulation, we found no difference in pressure laterality (-6.6{+/-}18.4 mmHg; p=0.185) or collateral flow rate (10{+/-}46 ml/min; p=0.421). TOF-MRA geometries can with signal intensity thresholding be matched to produce similar morphology and modelled cerebral perfusion pressure to CTA geometries. The modelled pressure drops over the collateral arteries were sensitive to the segmentation regardless of modality.
Brito-Pacheco, D. A.; Giannopoulos, P.; Reyes-Aldasoro, C. C.
Show abstract
In this work, the impact of outliers on the performance of machine learning and deep learning models is investigated, specifically for the case of histopathological images of colorectal cancer stained with Haematoxylin and Eosin. The evaluation of the impact is done through the systematic comparison of one machine learning model (Random Forests) and one deep learning model (ResNet-18). Both models were trained with the popular NCT-CRC-HE-VAL-100K dataset and tested on the CRC-HE-VAL-7K companion set. Then, a curation process was performed by analysing the divergence of patches based on chromatic, textural and topological features of the training set and removing outliers to repeat the training with a cleaned dataset. The results showed that machine learning models, can benefit more from improvements in the quality of data, than deep learning models. Further, the results suggest that deep learning models are more robust to outliers as, through the training process, the architectures can learn features other than those previously mentioned.
Harikumar, A.; Baker, B.; Amen, D.; Keator, D.; Calhoun, V. D.
Show abstract
Single photon emission computed tomography (SPECT) is a highly specialized imaging modality that enables measurement of regional cerebral perfusion and, in particular, resting cerebral blood flow (rCBF). Recent technological advances have improved SPECT quantification and reliability, making it increasingly useful for studying rCBF abnormalities and perfusion network alterations in psychiatric and neurological disorders. To characterize large scale functional organization in SPECT data, data driven decomposition methods such as independent component analysis (ICA) have been used to extract covarying perfusion patterns that map onto interpretable brain networks. Blind ICA provides a data driven approach to estimate these networks without strong prior assumptions. More recently, a hybrid approach that leverages spatial priors to guide a spatially constrained ICA (sc ICA) have been used to fully automate the ICA analysis while also providing participant-specific network estimates. While this has been reliably demonstrated in fMRI with the NeuroMark template, there is currently no comparable SPECT template. A SPECT template would enable automatic estimation of functional SPECT networks with participant-specific expressions that correspond across participants and studies. The current study introduces a new replicable NeuroMark SPECT template for estimating canonical perfusion covariance patterns (networks). We first identify replicable SPECT networks using blind ICA applied to two large sample SPECT datasets. We then demonstrate the use of the resulting template by applying sc-ICA to an independent schizophrenia dataset. In sum, this work presents and shares the first NeuroMark SPECT template and demonstrating its utility in an independent cohort, providing a scalable and robust framework for network-based analyses.
Hou, J.; Yi, X.; Li, C.; Li, J.; Cao, H.; Lu, Q.; Yu, X.
Show abstract
Predicting response to induction chemotherapy (IC) and overall survival (OS) is critical for optimizing treatment in patients with locally advanced nasopharyngeal carcinoma (LANPC). This study aimed to develop and validate a multi-task deep learning model integrating pretreatment MRI and whole slide images (WSIs) to predict IC response and OS in LANPC. Pretreatment MRI and WSIs from 404 patients with LANPC were retrospectively collected to construct a multi-task model (MoEMIL) for the simultaneous prediction of early IC response and OS. MoEMIL employed multi-instance learning to process WSIs, PyRadiomics and a convolutional neural network (ResNet50) to extract MRI features, and fused multimodal features through a multi-gate mixture-of-experts architecture. Clustering-constrained attention multiple instance learning and gradient-weighted class activation mapping were applied for visualization and interpretation. MoEMIL effectively stratified patients into good and poor IC response groups, achieving areas under the curve of 0.917, 0.869, and 0.801 in the train, validation, and test sets, respectively, and outperformed the deep learning radiomics model, the pathomics model and TNM staging. The model also stratified patients into high- and low-risk OS groups (P < 0.05). MoEMIL shows promise as a decision-support tool for early IC response prediction and prognostication in LANPC. Author SummaryWe have developed a deep learning model that integrates two types of medical images, including magnetic resonance imaging (MRI) and digital pathological slices, to simultaneously predict response to induction chemotherapy and prognosis in patients with locally advanced nasopharyngeal carcinoma. Current treatment decisions primarily rely on traditional tumor staging (TNM), which often fails to comprehensively reflect the complexity of the disease. Our model, named MoEMIL, was trained and tested on data from 404 patients across two hospitals and consistently outperformed both single-model approaches and TNM staging methods. By identifying patients who exhibit poor response to induction chemotherapy or higher prognostic risk, our tool can assist clinicians in achieving personalized treatment, enabling intensified management for high-risk patients and avoiding unnecessary side effects for low-risk patients. Additionally, we visualize the models reasoning process through heat map generation, which highlights the image regions exerting the greatest influence on prediction outcomes. This work represents a step toward more precise treatment for nasopharyngeal carcinoma; however, larger-scale prospective studies are required before the model can be integrated into routine clinical practice.
Smah, M. L.; Seale, A. C.; Rock, K. S.
Show abstract
Network-based epidemic models have been instrumental in understanding how contact structure shapes infectious disease dynamics, yet widely used frameworks such as Erd[o]s-Renyi, configuration-model, and stochastic block networks do not explicitly capture the combination of fully accessible (saturated) within-group interactions and constrained between-group connectivity characteristic of many real-world settings. Here, we introduce the Multi-Clique (MC) network model, a generative framework in which individuals are organised into fully connected cliques representing stable contact groups (e.g., households, classrooms, or workplaces), with a limited number of external connections governing inter-group transmission. Using stochastic susceptible-infectious-recovered (SIR) simulations on degree-matched networks, we compare epidemic dynamics on MC networks with those on classical random graph models. Despite having an identical mean degree, MC networks exhibit systematically distinct behaviour, including slower epidemic growth, reduced peak prevalence, increased fade-out probability, and delayed time to peak. These effects arise from rapid within but constrained between clique transmission, creating structural bottlenecks that standard models do not capture. The MC framework provides an interpretable, data-driven representation of recurrent contact structure, with parameters that map directly to observable quantities such as household and classroom sizes. By isolating the role of intergroup connectivity, the model offers a basis for evaluating targeted intervention strategies that reduce between-group mixing while preserving within-group interactions. Our results highlight the importance of explicitly representing the real-life clique-based network structure in epidemic models and suggest that classical degree-matched networks may systematically overestimate epidemic speed and intensity in structured populations.
Chandra, S.
Show abstract
Background. Pancreatic ductal adenocarcinoma (PDAC) has a five-year survival rate of approximately 12%, largely because it is typically diagnosed at an advanced stage. CT-based computational methods for early detection exist but rely on black-box deep learning or large texture feature sets without tissue-specific interpretability. Methods. We developed Virtual Spectral Decomposition (VSD), which applies six parameterized sigmoid functions S(HU) = 1/(1+exp(-alpha x (HU - mu))) to standard portal-venous CT, decomposing each pixel into tissue-specific response channels for fat (mu=-60), fluid (mu=10), parenchyma (mu=45), stroma (mu=75), vascular (mu=130), and calcification (mu=250). Dendritic Binary Gating identifies structural content per channel using morphological filtering, enabling co-firing analysis and lone firer identification. A 25-feature signature was extracted per patient. Three independent datasets were analyzed: NIH Pancreas-CT (n=78 healthy), Medical Segmentation Decathlon Task07 (n=281 PDAC, paired tumor/adjacent tissue), and CPTAC-PDA from The Cancer Imaging Archive (n=82, multi-institutional, with DICOM time point tags). The same six sigmoid parameters were used across all datasets without retraining. Results. VSD achieved AUC 0.943 for field effect detection (healthy vs cancer-adjacent parenchyma) and AUC 0.931 for patient-stratified tumor specification on MSD. On CPTAC-PDA, VSD achieved AUC 0.961 (6 features) and 0.979 (25 features) for distinguishing healthy from cancer-bearing pancreas on scans obtained prior to pathological diagnosis. All significant features replicated across datasets in the same direction: z_fat (d=-2.10, p=3.5e-27), z_fluid (d=-2.76, p=2.4e-38), fire_fat (d=+2.18, p=1.2e-28). Critically, VSD severity did not correlate with days-from-diagnosis (r=-0.008, p=0.944) across a range of day -1394 to day +249. Patient C3N-01375, scanned 3.8 years before pathological diagnosis, had VSD severity 1.87, well above the healthy mean of 0.94 +/- 0.33. The tissue transformation signature was temporally stable, indicating an early, persistent tissue state rather than a progressively worsening process. Conclusions. VSD with Dendritic Binary Gating detects a stable pancreatic tissue composition signature on standard CT that is present years before clinical diagnosis, validated across three independent datasets without parameter adjustment. The six sigmoid channels map to biologically meaningful tissue components through a fully transparent interpretability chain. The temporal stability of the signal implies a detection window of 3-7 years, consistent with known PanIN-3 microenvironment transformation timelines. VSD functions as a single-scan screening tool applicable to any abdominal CT performed during the pre-clinical window.
Perovic, M.; Mack, M. L.
Show abstract
Menstrual cycles are major biological events with extensive effects on the brain and cognition, experienced by half of the human population. To develop a comprehensive account of human cognition, it is necessary to successfully integrate and characterize menstrual cycle effects in cognitive science research. However, current approaches to menstrual cycle analysis suffer from low data resolution and are not well-equipped to capture the highly variable, non-linear changes in outcomes of interest across the cycle. We present a validated standardized method remedying these issues, demonstrate its utility using hormonal, behavioral, and neuroimaging data, and provide an open-source toolkit to facilitate its use.
Feng, Y.; Deng, K.; Guan, Y.
Show abstract
Gene networks (GNs) encode diverse molecular relationships and are central to interpreting cellular function and disease. The heterogeneity of interaction types has led to computational methods specialized for particular network contexts. Large language models (LLMs) offer a unified, language-based formulation of GN inference by leveraging biological knowledge from large-scale text corpora, yet their effectiveness remains sensitive to prompt design. Here, we introduce Gene-Relation Adaptive Soft Prompt (GRASP), a parameter-efficient and trainable framework that conditions inference on each gene pair through only three virtual tokens. Using factorized gene-specific and relation-aware components, GRASP learns to map each pair's biological context into compact soft prompts that combine pair-specific signals with shared interaction patterns. Across diverse GN inference tasks, GRASP consistently outperforms alternative prompting strategies. It also shows a stronger ability to recover unannotated interactions from synthetic negative sets, suggesting its capacity to identify biologically meaningful relationships beyond existing databases. Together, these results establish GRASP as a scalable and generalizable prompting framework for LLM-based GN inference.
Sivakumar, E.; Anand, A.
Show abstract
Computer vision and deep learning techniques, including convolutional neural networks (CNNs) and transformers, have increased the performance of medical image classification systems. However, training deep learning models using medical images is a challenging task that necessitates a substantial amount of annotated data. In this paper, we implement data augmentation strategies to tackle dataset imbalance in the VinDr-SpineXR dataset, which has a lower number of spine abnormality X-ray images compared to normal spine X-ray images. Geometric transformations and synthetic image generation using Generative Adversarial Networks are explored and applied to the abnormal classes of the dataset, and classifier performance is validated using VGG-16 and InceptionNet to identify the most effective augmentation technique. Additionally, we introduce a hybrid augmentation technique that addresses class imbalance, reduces computational overhead relative to a GAN-only approach, and achieves ~99% validation accuracy with both classifiers across all three case studies. Keywords: Data augmentation, Generative Adversarial Network, VGG-16, InceptionNet, Class imbalance, Computer vision, Spine X-ray, Radiology.